RankEval: Open Tool for Evaluation of Machine-Learned Ranking
نویسنده
چکیده
Recent research and applications for evaluation and quality estimation of Machine Translation require statistical measures for comparing machine-predicted ranking against gold sets annotated by humans. Additional to the existing practice of measuring segment-level correlation with Kendall tau, we propose using ranking metrics from the research field of Information Retrieval such as Mean Reciprocal Rank, Normalized Discounted Cumulative Gain and Expected Reciprocal Rank. These reward systems that predict correctly the highest ranked items than the one of lower ones. We present an open source tool ”RE” providing implementation of these metrics. It can be either run independently as a script supporting common formats or can be imported to any Python application.
منابع مشابه
COSTA MT Evaluation Tool: An Open Toolkit for Human Machine Translation Evaluation
Ahotly debated topic inmachine translation is human evaluation. On the one hand, it is extremely costly and time consuming; on the other, it is an important and unfortunately inevitable part of any system. This paper describes COSTAMT Evaluation Tool, an open stand-alone tool for human machine translation evaluation. It is a Java program that can be used to manually evaluate the quality of the ...
متن کاملارائه الگوریتمی مبتنی بر یادگیری جمعی به منظور یادگیری رتبهبندی در بازیابی اطلاعات
Learning to rank refers to machine learning techniques for training a model in a ranking task. Learning to rank has been shown to be useful in many applications of information retrieval, natural language processing, and data mining. Learning to rank can be described by two systems: a learning system and a ranking system. The learning system takes training data as input and constructs a ranking ...
متن کاملRanking journals: Could Google Scholar Metrics be an alternative to Journal Citation Reports and Scimago Journal Rank?
The launch of Google Scholar Metrics as a tool for assessing scientific journals may be serious competition for Thomson Reuters’ Journal Citation Reports, and for Scopus’ powered Scimago Journal Rank. , A review of these bibliometric journal evaluation products is performed. We compare their main characteristics from different approaches: coverage, indexing policies, search and visualization, b...
متن کاملTranscRater: a Tool for Automatic Speech Recognition Quality Estimation
We present TranscRater, an open-source tool for automatic speech recognition (ASR) quality estimation (QE). The tool allows users to perform ASR evaluation bypassing the need of reference transcripts and confidence information, which is common to current assessment protocols. TranscRater includes: i) methods to extract a variety of quality indicators from (signal, transcription) pairs and ii) m...
متن کاملAppraise: An Open-Source Toolkit for Manual Phrase-Based Evaluation of Translations
We describe a focused effort to investigate the performance of phrase-based, human evaluation of machine translation output achieving a high annotator agreement. We define phrase-based evaluation and describe the implementation of Appraise, a toolkit that supports the manual evaluation of machine translation results. Phrase ranking can be done using either a fine-grained six-way scoring scheme ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Prague Bull. Math. Linguistics
دوره 100 شماره
صفحات -
تاریخ انتشار 2013